Enhancing Parameter-Free Frank Wolfe with an Extra Subproblem
نویسندگان
چکیده
Aiming at convex optimization under structural constraints, this work introduces and analyzes a variant of the Frank Wolfe (FW) algorithm termed ExtraFW. The distinct feature ExtraFW is pair gradients leveraged per iteration, thanks to which decision variable updated in prediction-correction (PC) format. Relying on no problem dependent parameters step sizes, convergence rate for general problems shown be ${\cal O}(\frac{1}{k})$, optimal sense matching lower bound number solved FW subproblems. However, merit its faster O}\big(\frac{1}{k^2} \big)$ class machine learning problems. Compared with other parameter-free variants that have rates same problems, has improved fine-grained analysis PC update. Numerical tests binary classification different sparsity-promoting constraints demonstrate empirical performance significantly better than FW, even Nesterov's accelerated gradient certain datasets. For matrix completion, enjoys smaller optimality gap, rank FW.
منابع مشابه
Frank-Wolfe with Subsampling Oracle
We analyze two novel randomized variants of the Frank-Wolfe (FW) or conditional gradient algorithm. While classical FW algorithms require solving a linear minimization problem over the domain at each iteration, the proposedmethod only requires to solve a linear minimization problem over a small subset of the original domain. The first algorithm that we propose is a randomized variant of the ori...
متن کاملRevisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization
We provide stronger and more general primal-dual convergence results for FrankWolfe-type algorithms (a.k.a. conditional gradient) for constrained convex optimization, enabled by a simple framework of duality gap certificates. Our analysis also holds if the linear subproblems are only solved approximately (as well as if the gradients are inexact), and is proven to be worst-case optimal in the sp...
متن کاملCombining Progressive Hedging with a Frank-wolfe
We present a new primal-dual algorithm for computing the value of the Lagrangian 6 dual of a stochastic mixed-integer program (SMIP) formed by relaxing its nonanticipativity con7 straints. The algorithm relies on the well-known progressive hedging method, but unlike previous 8 progressive hedging approaches for SMIP, our algorithm can be shown to converge to the optimal 9 Lagrangian dual value....
متن کاملLearning Infinite RBMs with Frank-Wolfe
In this work, we propose an infinite restricted Boltzmann machine (RBM), whose maximum likelihood estimation (MLE) corresponds to a constrained convex optimization. We consider the Frank-Wolfe algorithm to solve the program, which provides a sparse solution that can be interpreted as inserting a hidden unit at each iteration, so that the optimization process takes the form of a sequence of fini...
متن کاملBandit Optimization with Upper-Confidence Frank-Wolfe
We consider the problem of bandit optimization, inspired by stochastic optimization and online learning problems with bandit feedback. In this problem, the objective is to minimize a global loss function of all the actions, not necessarily a cumulative loss. This framework allows us to study a very general class of problems, with applications in statistics, machine learning, and other fields. T...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2021
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v35i9.17012